Retrofitting Decision Tree Classifiers Using Kernel Density Estimation
نویسندگان
چکیده
A ]IOVC1 mdl)d for cxnnbining dccisio]l trcws a]ld kcn)d dmlsity cstlimators i s ]woposcd. Sta]ldard classification]) tmcs, or class prob al)ility trms, ]movidc piuwwisc constant estimates of class posterior ]mobabilitlics. KcrI)C1 dmlsity estimators can ])rovidc smooth ]Io]l-])alalllct]ic estimates of class probaliitics, l)ut scale ])oorly as the dilncl]sionality o f tllc ])roblcm illcrcascs. ‘J’his palm dismsscs a l)ylnid sclIcIIlc whicl] uscx decision trees to find t,llc rclcwal)t s t ructure ill IIig]l-(li]llc]lsio]lal classification lnoblmns and tllml uses local km)cl density esti]nat,cs to fit Slnootl) ]mobability cstilllates within t h i s structure. IJxlmrilnclltal mmlts 011 simulated da ta indicate that tl)c mctllod ~wovidcs sub stalltial i]nlmwmncl]t over trees or clmlsity ]I]etllods alo]lc for ccrtaill c lasses of ]moblmns. ‘1’llc I)al)cr briefly discusses various cxtcllsiolls of tllc basic a])lnoacll and t,llc types of a]q)]icatiou for wllicll tllc ]nctllod is Lest suitd.
منابع مشابه
Bagging classifiers based on Kernel density estimators
A lot of research is being conducted on combining classification rules (classifiers) to produce a single one, known as an ensemble, which in general is more accurate than the individual classifiers making up the ensemble. Two popular methods for creating ensembles are Bagging introduced by Breiman, (1996) and, AdaBoosting by Freund and Schapire (1996). These methods rely on resampling technique...
متن کاملTransductive Reliability Estimation for Kernel Based Classifiers
Estimating the reliability of individual classifications is very important in several applications such as medical diagnosis. Recently, the transductive approach to reliability estimation has been proved to be very efficient when used with several machine learning classifiers, such as Naive Bayes and decision trees. However, the efficiency of the transductive approach for state-of-the art kerne...
متن کاملEstimating Continuous Distributions in Bayesian Classifiers
When modeling a probability distribution with a Bayesian network, we are faced with the problem of how to handle continuous vari ables. Most previous work has either solved the problem by discretizing, or assumed that the data are generated by a single Gaussian. In this paper we abandon the normality as sumption and instead use statistical methods for nonparametric density estimation. For a n...
متن کاملRetro tting Decision Tree Classi ers Using Kernel
A novel method for combining decision trees and kernel density estimators is proposed. Standard classiication trees, or class probability trees, provide piecewise constant estimates of class posterior probabilities. Kernel density estimators can provide smooth non-parametric estimates of class probabilities , but scale poorly as the dimensional-ity of the problem increases. This paper discusses...
متن کاملParallel computation of kernel density estimates classifiers and their ensembles
Nonparametric supervised classifiers are interesting because they do not require distributional assumptions for the class conditional density, such as normality or equal covariance. However their use is not widespread because it takes a lot of time to compute them due to the intensive use of the available data. On the other hand bundling classifiers to produce a single one, known as an ensemble...
متن کامل